Properties of bagged nearest neighbour classifiers
نویسندگان
چکیده
منابع مشابه
Properties of bagged nearest neighbour classifiers
It is shown that bagging, a computationally intensive method, asymptotically improves the performance of nearest neighbour classifiers provided that the resample size is less than 69% of the actual sample size, in the case of with-replacement bagging, or less than 50% of the sample size, for without-replacement bagging. However, for larger sampling fractions there is no asymptotic difference be...
متن کاملk-Nearest Neighbour Classifiers
Perhaps the most straightforward classifier in the arsenal or machine learning techniques is the Nearest Neighbour Classifier – classification is achieved by identifying the nearest neighbours to a query example and using those neighbours to determine the class of the query. This approach to classification is of particular importance today because issues of poor run-time performance is not such...
متن کاملOptimal weighted nearest neighbour classifiers
We derive an asymptotic expansion for the excess risk (regret) of a weighted nearest-neighbour classifier. This allows us to find the asymptotically optimal vector of nonnegative weights, which has a rather simple form. We show that the ratio of the regret of this classifier to that of an unweighted k-nearest neighbour classifier depends asymptotically only on the dimension d of the feature vec...
متن کاملTo “ Optimal Weighted Nearest Neighbour Classifiers ”
Completion of the proof of Theorem 1. Step 6: To show (6.5), which gives the Radon–Nikodym derivative of the restriction of the distribution of X(i) − x to a small ball about the origin Recall that Bδ(u) = {y ∈ Rd : ‖y − u‖ ≤ δ}, and that νd denotes ddimensional Lebesgue measure. For a Borel subset A of Rd, let N(A) = ∑n i=1 1{Xi∈A}. It follows from the hypothesis (A.2) that for x ∈ S n with n ...
متن کاملNearest-neighbour classifiers in natural scene analysis
It is now well-established that k nearest-neighbour classi"ers o!er a quick and reliable method of data classi"cation. In this paper we extend the basic de"nition of the standard k nearest-neighbour algorithm to include the ability to resolve con#icts when the highest number of nearest neighbours are found for more than one training class (model-1). We also propose model-2 of nearest-neighbour ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Journal of the Royal Statistical Society: Series B (Statistical Methodology)
سال: 2005
ISSN: 1369-7412,1467-9868
DOI: 10.1111/j.1467-9868.2005.00506.x